Hadoop Mapreduce OpenCL Plugin
نویسنده
چکیده
Modern systems generates huge amounts of information right from areas like finance, telematics, healthcare, IOT devices to name a few, the modern day computing frameworks like Mapreduce needs an ever increasing amount of computing power to sort, arrange and generate insights from the data. This project is an attempt to harness the power of heterogeneous computing, more specifically take benefit of parallelism offered by modern day GPU‟s and accelerate data processing of Mappers and Reducers in the Mapreduce framework. In this regard, AMD‟s opensource APARAPI library was used as the foundation to develop a simple to use heterogeneous framework. The framework uses both the CPU and GPU to perform processing by translating sections of Mapreduce Java bytecodes to OpenCL and executing the same in parallel in AMD GPU‟s. Typical Mapreduce tasks were then run in this framework with various types of workloads and while no perceptible improvement in speedups were noticed the framework shows excellent promise in computing complicated parallel data in upcoming high performance computing workloads.
منابع مشابه
Teaching parallel programming to undergrads with hands-on experience
This paper describes the didactic concept, the content and the lessons learned of a lecture on parallel programming for undergraduate students held during summer term 2013. The course’s focus was on providing hands-on experience, hence students were programming on real life codes using an actual HPC cluster. The lecture’s aim was to provide an in-depth understanding of each parallel programming...
متن کاملA Research of MapReduce with GPU Acceleration
MapReduce is an efficient distributed computing model on large data sets. The data processing is fully distributed on huge amount of nodes, and a MapReduce cluster is of highly scalable. However, single-node performance is gradually to be a bottleneck in computeintensive jobs, which makes it difficult to extend the MapReduce model to wider application fields such as largescale image processing ...
متن کاملHadoop+Aparapi: Making heterogenous MapReduce programming easier
Lately, programmers have started to take advantage of GPU capabilities of cloud-based machines. Using the GPUs can decrease the number of nodes required to perform the computation by increasing the productivity per node. We combine Hadoop, a widely-used MapReduce framework, with Aparapi, a new Java-to-OpenCL conversion tool from AMD. We propose an easy-to-use API which allows easy implementatio...
متن کاملAdaptive Dynamic Data Placement Algorithm for Hadoop in Heterogeneous Environments
Hadoop MapReduce framework is an important distributed processing model for large-scale data intensive applications. The current Hadoop and the existing Hadoop distributed file system’s rack-aware data placement strategy in MapReduce in the homogeneous Hadoop cluster assume that each node in a cluster has the same computing capacity and a same workload is assigned to each node. Default Hadoop d...
متن کاملHigh Level Programming for Heterogeneous Architectures
This work presents an effort to bridge the gap between abstract high level programming and OpenCL by extending an existing high level Java programming framework (APARAPI), based on OpenCL, so that it can be used to program FPGAs at a high level of abstraction and increased ease of programmability. We run several real world algorithms to assess the performance of the framework on both a low end ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2015